151 research outputs found
Personalized PageRank with Node-dependent Restart
Personalized PageRank is an algorithm to classify the improtance of web pages
on a user-dependent basis. We introduce two generalizations of Personalized
PageRank with node-dependent restart. The first generalization is based on the
proportion of visits to nodes before the restart, whereas the second
generalization is based on the probability of visited node just before the
restart. In the original case of constant restart probability, the two measures
coincide. We discuss interesting particular cases of restart probabilities and
restart distributions. We show that the both generalizations of Personalized
PageRank have an elegant expression connecting the so-called direct and reverse
Personalized PageRanks that yield a symmetry property of these Personalized
PageRanks
Analysis of Relaxation Time in Random Walk with Jumps
We study the relaxation time in the random walk with jumps. The random walk
with jumps combines random walk based sampling with uniform node sampling and
improves the performance of network analysis and learning tasks. We derive
various conditions under which the relaxation time decreases with the
introduction of jumps.Comment: 13 page
Bayesian Inference of Online Social Network Statistics via Lightweight Random Walk Crawls
Online social networks (OSN) contain extensive amount of information about
the underlying society that is yet to be explored. One of the most feasible
technique to fetch information from OSN, crawling through Application
Programming Interface (API) requests, poses serious concerns over the the
guarantees of the estimates. In this work, we focus on making reliable
statistical inference with limited API crawls. Based on regenerative properties
of the random walks, we propose an unbiased estimator for the aggregated sum of
functions over edges and proved the connection between variance of the
estimator and spectral gap. In order to facilitate Bayesian inference on the
true value of the estimator, we derive the approximate posterior distribution
of the estimate. Later the proposed ideas are validated with numerical
experiments on inference problems in real-world networks
Bidirectional PageRank Estimation: From Average-Case to Worst-Case
We present a new algorithm for estimating the Personalized PageRank (PPR)
between a source and target node on undirected graphs, with sublinear
running-time guarantees over the worst-case choice of source and target nodes.
Our work builds on a recent line of work on bidirectional estimators for PPR,
which obtained sublinear running-time guarantees but in an average-case sense,
for a uniformly random choice of target node. Crucially, we show how the
reversibility of random walks on undirected networks can be exploited to
convert average-case to worst-case guarantees. While past bidirectional methods
combine forward random walks with reverse local pushes, our algorithm combines
forward local pushes with reverse random walks. We also discuss how to modify
our methods to estimate random-walk probabilities for any length distribution,
thereby obtaining fast algorithms for estimating general graph diffusions,
including the heat kernel, on undirected networks.Comment: Workshop on Algorithms and Models for the Web-Graph (WAW) 201
Monte Carlo methods in PageRank computation: When one iteration is sufficient
PageRank is one of the principle criteria according to which Google ranks Web pages. PageRank can be interpreted as a frequency of visiting a Web page by a random surfer and thus it reflects the popularity of a Web page. Google computes the PageRank using the power iteration method which requires about one week of intensive computations. In the present work we propose and analyze Monte Carlo type methods for the PageRank computation. There are several advantages of the probabilistic Monte Carlo methods over the deterministic power iteration method: Monte Carlo methods provide good estimation of the PageRank for relatively important pages already after one iteration; Monte Carlo methods have natural parallel implementation; and finally, Monte Carlo methods allow to perform continuous update of the PageRank as the structure of the Web changes
Stability of constant retrial rate systems with NBU input*
We study the stability of a single-server retrial queueing system with constant retrial rate, general input and service processes. First, we present a review of some relevant recent results related to the stability criteria of similar systems. Sufficient stability conditions were obtained by Avrachenkov and Morozov (2014), which hold for a rather general retrial system. However, only in the case of Poisson input is an explicit expression provided; otherwise one has to rely on simulation. On the other hand, the stability criteria derived by Lillo (1996) can be easily computed but only hold for the case of exponential service times. We present new sufficient stability conditions, which are less tight than the ones obtained by Avrachenkov and Morozov (2010), but have an analytical expression under rather general assumptions. A key assumption is that interarrival times belongs to the class of new better than used (NBU) distributions. We illustrate the accuracy of the condition based on this assumption (in comparison with known conditions when possible) for a number of non-exponential distributions
Multivariate polynomial perturbations of algebraic equations
AbstractIn this note we study multivariate perturbations of algebraic equations. In general, it is not possible to represent the perturbed solution as a Puiseux-type power series in a connected neighborhood. For the case of two perturbation parameters we provide a sufficient condition that guarantees such a representation. Then, we extend this result to the case of more than two perturbation parameters. We motivate our study by the perturbation analysis of a weighted random walk on the Web Graph. In an instance of the latter the stationary distribution of the weighted random walk, the so-called Weighted PageRank, may depend on two (or more) perturbation parameters in a manner that illustrates our theoretical development
Convergence of trajectories and optimal buffer sizing for AIMD congestion control
We study the interaction between the AIMD (Additive Increase Multiplicative Decrease) multi-socket congestion control and a bottleneck router with Drop Tail buffer. We consider the problem in the framework of deterministic hybrid models. First, we show that trajectories always converge to limiting cycles. We characterize the cycles. Necessary and sufficient conditions for the absence of multiple jumps in the same cycle are obtained. Then, we propose an analytical framework for the optimal choice of the router buffer size. We formulate this problem as a multi-criteria optimization problem, in which the Lagrange function corresponds to a linear combination of the average goodput and the average delay in the queue. Our analytical results are confirmed by simulations performed with MATLAB Simulink
- …